Numerical Issues Involved in Inverting Hessian Matrices

نویسندگان

  • Jeff Gill
  • Gary King
  • Michael P. McDonald
چکیده

In the social sciences, researchers typically assume the accuracy of generalized linear models by using an asymptotic normal approximation to the likelihood function or, occasionally, by using the full posterior distribution. Thus, for standard maximum likelihood analyses, only point estimates and the variance at the maximum are normally seen as necessary. For Bayesian posterior analysis, the maximum and variance provide a useful first approximation (but see Chapter 4 for an alternative). Unfortunately, although the negative of the Hessian (the matrix of second derivatives of the posterior with respect to the parameters and named for its inventor in slightly different context, German mathematician Ludwig Hesse) must be positive definite and hence invertible so as to compute the variance matrix, invertible Hessians do not exist for some combinations of datasets and models, so statistical procedures sometimes fail for this reason before completion. Indeed, receiving a computer-generated "Hessian not invertible" message (because of singularity or nonpositive definiteness) rather than a set of statistical results is a frustrating but common occurrence in applied quantitative research. It even occurs with regularity during many Monte Carlo experiments where the investigator is drawing data from a known statistical model, due to machine effects. The Hessian can be noninvertible for both computational reasons and data reasons. Inaccurate implementation of the likelihood function (see Chapters 2 and 3), inaccurate derivative methods (see Chapter 8), or other inappropriate choices in optimization algorithms can yield noninvertible Hessians. Where these inaccuracies cause problems with Hessians, we recommend addressing these inaccuracies directly. If these methods aren't feasible, or don't work, which often happens, we provide an innovative new library for doing generalized inverses . Moreover, when a Hessian is not invertible for data reasons, no computational trick can make it invertible, given the model and data chosen, because the desired inverse does

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Hessian-Based Method for Uncertainty Quantification in Global Ocean State Estimation

Derivative-based methods are developed for uncertainty quantification (UQ) in largescale ocean state estimation. The estimation system is based on the adjoint method for solving a least-squares optimization problem, whereby the state-of-the-art MIT general circulation model (MITgcm) is fit to observations. The UQ framework is applied to quantify Drake Passage transport uncertainties in a global...

متن کامل

Matrix Probing and its Conditioning

When a matrix A with n columns is known to be well-approximated by a linear combination of basis matrices B1, . . . , Bp, we can apply A to a random vector and solve a linear system to recover this linear combination. The same technique can be used to obtain an approximation to A−1. A basic question is whether this linear system is well-conditioned. This is important for two reasons: a well-con...

متن کامل

Convergence of Rump’s Method for Inverting Arbitrarily Ill-Conditioned Matrices

In this paper, the problem of inverting regular matrices with arbitrarily large condition number is treated in double precision defined by IEEE 754 floating point standard. In about 1984, Rump derived a method for inverting arbitrarily ill-conditioned matrices. The method requires the possibility to calculate a dot product in higher precision. Rump’s method is of theoretical interest. Rump made...

متن کامل

Variable-storage quasi-Newton operators as inverse forecast/analysis error covariance matrices in variational data assimilation

Two approximations of the Hessian matrix as limited-memory operators are built from the limited-memoryBFGS inverse Hessian approximationprovided by the minimization code, in view of the speci cation of the inverse analysis/forecast error covariance matrix in variational data assimilation. Some numerical experiments and theoretical considerations lead to reject the limited-memory DFP Hessian app...

متن کامل

Approximate inversion of the wave-equation Hessian via randomized matrix probing

We present a method for approximately inverting the Hessian of full waveform inversion as a dip-dependent and scaledependent amplitude correction. The terms in the expansion of this correction are determined by least-squares fitting from a handful of applications of the Hessian to random models — a procedure called matrix probing. We show numerical indications that randomness is important for g...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2005